Video Depth Anything
Annolid bundles the Video-Depth-Anything models to estimate per-frame depth maps from a video.
Run in the GUI
Open a video (File → Open Video).
(Optional) Tune defaults in View → Depth Settings… (encoder, resolution, target FPS, streaming, and which outputs to save).
Start processing via View → Video Depth Anything….
Outputs are written next to the video by default, in a folder named after the video stem (e.g. videos/mice.mp4 → videos/mice/).
Download checkpoints (optional)
Checkpoints live under annolid/depth/checkpoints and are auto-downloaded when you run the tool. To prefetch them:
python -m annolid.depth.download_weights --list
python -m annolid.depth.download_weights --model video_depth_anything_vits
Set HF_HUB_TOKEN if you need authentication to download from Hugging Face.
Run from Python
from annolid.depth import run_video_depth_anything
run_video_depth_anything(
input_video="videos/mice.mp4",
output_dir="outputs/mice_depth",
encoder="vits",
streaming=True,
target_fps=15,
save_depth_video=True,
)
Outputs
depth.ndjson(always): one JSON record per processed frame, with a base64-encodeduint16PNG depth map and a per-frame min/maxscale.<video_stem>_vis.mp4(optional): saved whensave_depth_video=True.depth_frames/andpoint_clouds/(optional): saved when enabled in settings.